SdAE: Self-distillated Masked Autoencoder

نویسندگان

چکیده

With the development of generative-based self-supervised learning (SSL) approaches like BeiT and MAE, how to learn good representations by masking random patches input image reconstructing missing information has grown in concern. However, PeCo need a “pre-pretraining” stage produce discrete codebooks for masked representing. MAE does not require pre-training codebook process, but setting pixels as reconstruction targets may introduce an optimization gap between downstream tasks that quality always lead high descriptive capability model. Considering above issues, this paper, we propose simple Self-distillated AutoEncoder network, namely SdAE. SdAE consists student branch using encoder-decoder structure reconstruct information, teacher producing latent representation tokens. We also analyze build views from perspective bottleneck. After that, multi-fold strategy provide multiple with balanced boosting performance, which can reduce computational complexity. Our approach generalizes well: only 300 epochs pre-training, vanilla ViT-Base model achieves 84.1% fine-tuning accuracy on ImageNet-1 k classification, 48.6 mIOU ADE20K segmentation, 48.9 mAP COCO detection surpasses other methods considerable margin. Code is available at https://github.com/AbrahamYabo/SdAE .

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

MADE: Masked Autoencoder for Distribution Estimation

There has been a lot of recent interest in designing neural network models to estimate a distribution from a set of examples. We introduce a simple modification for autoencoder neural networks that yields powerful generative models. Our method masks the autoencoder’s parameters to respect autoregressive constraints: each input is reconstructed only from previous inputs in a given ordering. Cons...

متن کامل

Autoencoder Trees

We discuss an autoencoder model in which the encoding and decoding functions are implemented by decision trees. We use the soft decision tree where internal nodes realize soft multivariate splits given by a gating function and the overall output is the average of all leaves weighted by the gating values on their path. The encoder tree takes the input and generates a lower dimensional representa...

متن کامل

AutoEncoder by Forest

Auto-encoding is an important task which is typically realized by deep neural networks (DNNs) such as convolutional neural networks (CNN). In this paper, we propose EncoderForest (abbrv. eForest), the first tree ensemble based auto-encoder. We present a procedure for enabling forests to do backward reconstruction by utilizing the equivalent classes defined by decision paths of the trees, and de...

متن کامل

Cutting Recursive Autoencoder Trees

Deep Learning models enjoy considerable success in Natural Language Processing. While deep architectures produce useful representations that lead to improvements in various tasks, they are often difficult to interpret. This makes the analysis of learned structures particularly difficult. In this paper, we rely on empirical tests to see whether a particular structure makes sense. We present an a...

متن کامل

Autoencoder for words

This paper presents a training method that encodes each word into a different vector in semantic space and its relation to low entropy coding. Elman network is employed in the method to process word sequences from literary works. The trained codes possess reduced entropy and are used in ranking, indexing, and categorizing literary works. A modification of the method to train the multi-vector fo...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2022

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-031-20056-4_7